53 research outputs found
Virtual to Real Reinforcement Learning for Autonomous Driving
Reinforcement learning is considered as a promising direction for driving
policy learning. However, training autonomous driving vehicle with
reinforcement learning in real environment involves non-affordable
trial-and-error. It is more desirable to first train in a virtual environment
and then transfer to the real environment. In this paper, we propose a novel
realistic translation network to make model trained in virtual environment be
workable in real world. The proposed network can convert non-realistic virtual
image input into a realistic one with similar scene structure. Given realistic
frames as input, driving policy trained by reinforcement learning can nicely
adapt to real world driving. Experiments show that our proposed virtual to real
(VR) reinforcement learning (RL) works pretty well. To our knowledge, this is
the first successful case of driving policy trained by reinforcement learning
that can adapt to real world driving data
Unsupervised Adaptation from Repeated Traversals for Autonomous Driving
For a self-driving car to operate reliably, its perceptual system must
generalize to the end-user's environment -- ideally without additional
annotation efforts. One potential solution is to leverage unlabeled data (e.g.,
unlabeled LiDAR point clouds) collected from the end-users' environments (i.e.
target domain) to adapt the system to the difference between training and
testing environments. While extensive research has been done on such an
unsupervised domain adaptation problem, one fundamental problem lingers: there
is no reliable signal in the target domain to supervise the adaptation process.
To overcome this issue we observe that it is easy to collect unsupervised data
from multiple traversals of repeated routes. While different from conventional
unsupervised domain adaptation, this assumption is extremely realistic since
many drivers share the same roads. We show that this simple additional
assumption is sufficient to obtain a potent signal that allows us to perform
iterative self-training of 3D object detectors on the target domain.
Concretely, we generate pseudo-labels with the out-of-domain detector but
reduce false positives by removing detections of supposedly mobile objects that
are persistent across traversals. Further, we reduce false negatives by
encouraging predictions in regions that are not persistent. We experiment with
our approach on two large-scale driving datasets and show remarkable
improvement in 3D object detection of cars, pedestrians, and cyclists, bringing
us a step closer to generalizable autonomous driving.Comment: Accepted by NeurIPS 2022. Code is available at
https://github.com/YurongYou/Rote-D
- …